perm filename MENTAL[F75,JMC]3 blob sn#202281 filedate 1976-02-13 generic text, type C, neo UTF8
COMMENT ⊗   VALID 00010 PAGES
C REC  PAGE   DESCRIPTION
C00001 00001
C00002 00002	.SPACING 1
C00003 00003	.COMPACT INDENT 5
C00012 00004	WHY ASCRIBE MENTAL QUALITIES?
C00022 00005	TWO METHODS OF DEFINITION
C00043 00006	EXAMPLES OF SYSTEMS WITH MENTAL QUALITIES
C00076 00007	"GLOSSARY" OF MENTAL QUALITIES
C00091 00008	OTHER VIEWS ABOUT MIND
C00092 00009	.PORTION NOTES
C00101 00010	REFERENCES
C00102 ENDMK
C⊗;
.SPACING 1;
.TURN ON "{";
.DEVICE XGP;
.FONT 1 "BASL30"; TURN ON "%";
.FONT 2 "BASI30";
.FONT 3 "basb30";
.FONT 4 "SUB";
.FONT 5 "SUP";
.at "⊗"word" "	⊂ ("%2word%* ") ⊃;
.COUNT ITEM
.COUNT SUBITEM IN ITEM
.COUNT NOTE inline
.AT "~" TEXT "~"; ⊂
.SEND NOTES ⊂
{NOTE!}. TEXT

.⊃;
.⊃;
.AT "$" ⊂ NEXT NOTE; ("%5"&NOTE!&"%*");⊃;
.AT "#" ⊂NEXT ITEM;(ITEM!);⊃;
.AT "&" ⊂NEXT SUBITEM;ONCE INDENT 10;(ITEM!&"."&SUBITEM!);⊃;
.PAGE FRAME 53 HIGH 80 WIDE;
.SECNUM←0
.MACRO NS(A)  ⊂ SECNUM←SECNUM+1
.SKIP TO COLUMN 1
.ONCE CENTER;
%3{SECNUM}.  A%*
.  ⊃
.EVERY HEADING(PARTIAL DRAFT, ,{DATE})
.EVERY FOOTING(,{PAGE},)
.COMPACT; INDENT 5;
.NEXT PAGE
.PLACE TEXT
.SELECT 1
.ONCE CENTER
(this draft compiled at {TIME} on {DATE})
.SKIP 2
.ONCE CENTER
ASCRIBING MENTAL QUALITIES TO MACHINES


	To ascribe certain %2beliefs%1, %2knowledge%1, %2free will%1,
%2intentions%1,  %2consciousness%1,  %2abilities%1  or %2wants%1 to a
machine or computer program is %3legitimate%1  whenever  such  ascription
expresses  the  same  information  about  the machine that it expresses
about a person.     It  is  %3useful%1  when  the  ascription  helps  us
understand the structure of the machine, its past or future behavior,
or how to repair or improve it.
It is perhaps never %3logically required%1 even for humans,
but a practical theory of the behavior of machines
or humans may require mental qualities or qualities isomorphic to them.
Theories of belief,  knowledge  and
wanting can be constructed for machines in a simpler setting than for
humans and later applied to humans.    Ascription of mental qualities
is %3most straightforward%1 for machines of known structure such as
thermostats and computer operating systems, but is %3most useful%1  when
applied  to  machines  whose  structure  is  very incompletely known.

	The above  views come from  work in  artificial intelligence$
(abbreviated  AI).  They can  be taken as asserting that
many of the philosophical problems  of mind take a practical form  as
soon  as one  takes  seriously  the idea  of  making machines  behave
intelligently.  In particular, AI raises for machines two issues that
have heretofore been considered only in connection with people.

	First, in designing intelligent programs  and looking at them
from  the outside  we need  to determine  the conditions  under which
specific mental and volitional terms  are applicable.  We can  exemplify
these  problems by  asking  when  it is  legitimate  to  say about  a
machine, %2" It knows I want a reservation to Boston, and it can give
it to me, but it won't"%1.

~We don't claim that the work in artificial intelligence has yet shown
how to reach human-level intellectual performance.
Our approach to the AI problem involves identifying the intellectual
mechanisms required for problem solving and describing them precisely,
and therefore we are at the end of the philosophical
spectrum that requires everything to be formalized in mathematical logic.
It  is sometimes  said  that  one
studies philosophy  in order to advance beyond  one's untutored naive
world-view, but unfortunately for  artificial intelligence, no-one has
yet been able to give a precise description of even a naive world-view.
~
	Second, when we  want a %3generally intelligent%1$ computer
program, we must build into it a %3general view%1 of what the world
is like with  especial attention to facts  about how the  information
required to solve problems is to be  obtained and used.  Thus we must
provide   it   with   some   kind   of   %2metaphysics%1 (general
world-view)  and %2epistemology%1 (theory of knowledge) however naive.

~Present AI programs operate in limited domains, e.g. play particular
games, prove theorems in a particular logical system, or understand
natural language sentences covering a particular subject matter and
with other semantic restrictions.  General intelligence will require
general models of situations changing in time, actors with goals
and strategies for achieving them, and knowledge about how information
can be obtained.~

	As  much  as  possible,  we  will  ascribe  mental  qualities
separately from  each other instead of bundling  them  in a
concept of mind.   This is necessary,  because present machines  have
rather   varied  little   minds;  the   mental  qualities   that  can
legitimately  be ascribed to them are few  and differ from machine to
machine.   We  will  not  even try  to  meet  objections  like,
%2"Unless it also  does X, it is illegitimate to  speak of its having
mental qualities."%1

	Machines  as  simple  as thermostats  can  be  said  to  have
beliefs,  and having  beliefs seems  to be  a characteristic  of most
machines  capable of  problem  solving  performance.    However,  the
machines mankind has so far found  it useful to construct rarely have
beliefs  about beliefs.    (Beliefs about  beliefs will be  needed by
computer programs to reason about what  knowledge they
lack and where  to get it).  Other mental qualities,  such as love and
hate, seem peculiar  to human-like motivational  structures and  will
not  be required  for  intelligent behavior,  but  we could  probably
program computers to exhibit them if we wanted to, because our common
sense notions about them  translate readily into certain program  and
data  structures.   Still  other  mental  qualities, e.g.  humor  and
appreciation  of beauty,  seem much  harder to  model.  While  we will be
quite liberal  in  ascribing ⊗some  mental  qualities even  to  rather
primitive machines, we will try to be conservative in our criteria
for ascribing any ⊗particular quality.

	The successive sections of this paper will give philosophical
and AI reasons for ascribing beliefs to machines, two new forms of
definition that seem necessary for defining mental qualities and
examples of their use, examples of systems to which mental
qualities are ascribed, some first attempts at defining a variety
of mental qualities, some criticisms of other views on mental
qualities, notes, and references.

.SKIP TO COLUMN 1
.ONCE CENTER
WHY ASCRIBE MENTAL QUALITIES?


	%3Why  should we want to  ascribe beliefs to  machines  at all?%1
This is the opposite question to that involved in the controversy over
%2reductionism%1.  Instead of asking how mental qualities can be
%3reduced%1 to physical ones, we ask how to %3ascribe%1 mental qualities
to physical systems.  In our opinion, this question is more natural
and may lead to better answers to the questions of reductionism.

	To put  the issue  sharply, consider  a computer program  for
which  we posess complete listings.   The behavior of  the program in
any environment is determined from  the structure of the program  and
can be  found out  by simulating  the action of  the program  and the
environment  without  having  to deal  with  any  concept  of belief.
Nevertheless, there  are  several reasons  for  ascribing belief  and
other mental qualities:

.ITEM←0;
	#. Although we may know the program, its state at a given moment
is usually not directly observable, and the conclusions we can draw
about its current state may be more readily expressed by ascribing
certain beliefs or wants than in any other way.

	#. Even if we can simulate the interaction of our program
with its environment using another more comprehensive program, the
simulation may be a billion times too slow.  We also
may not have the initial conditions of the environment or its laws of
motion in  a suitable  form, whereas  it may  be feasible  to make  a
prediction of  the effects of the  beliefs we ascribe to  the program
without any computer at all.

	#. Ascribing beliefs may allow deriving general statements
aobut the program's behavior that could not be obtained from any
finite number of simulations.

	#.   The belief and goal structures we ascribe to the program
may be easier to understand than the details of  program as expressed
in its listing.

	#. The belief and goal structure is likely to be close to the
structure the designer  of the  program had in  mind, and  it may  be
easier to debug the program in terms  of this structure than directly
from  the listing.   In  fact, it  is often  possible for  someone to
correct a fault by reasoning  in general terms about the  information
in a program or machine, diagnosing  what is wrong as a false belief,
and   looking  at  the  details  of   the  program  or  machine  only
sufficiently to determine how  the false belief is represented  and
what mechanism caused it to arise.$

	All the above reasons for ascribing beliefs are epistemological.
i.e. ascribing beliefs is needed to adapt to limitations on our
ability to acquire knowledge, use it for prediction, and establish
generalizations in terms of the elementary structure of the program.
Perhaps this is the general reason for ascribing higher levels of
organization to systems.

~	This  kind  of  teleological  analysis  is  often  useful  in
understanding natural organisms as well as machines.  Here  evolution
takes  the place  of  design and  we  often understand  the  function
performed by  an organ before we  understand its detailed physiology.
Teleological analysis  is  applicable to  psychological  and  social
phenomena in  so far as  these are designed  or have been  subject to
selection.

	However,  teleological analysis fails when applied to aspects
of nature which  have neither been  designed nor produced by  natural
selection from a population.   Much medieval science was based on the
Judeo-Christian-Moslem religious hypothesis that  the details of  the
world were designed by God  for the benefit of man.   The strong form
of  this hypothesis was abandoned  at the time of  Galileo and Newton
but occasionally recurs.   Barry Commoner's (1972) axiom of  ecology
"Nature knows best"  seems to be mistakenly based on  the notion that
nature  as a  whole  is the  result of  an evolutionary  process that
selected the "best nature".~

	Computers give rise to numerous examples of building a higher
structure on the basis of a lower and conducting subsequent analyses
using the higher structure.  The geometry of the electric fields in
a transistor and its chemical composition give rise to its properties
as an electric circuit element.
Transistors are combined in small circuits and powered in standard
ways to make logical elements such as ANDs, ORs, NOTs and flip-flops.
Computers are designed with these logical elements
to obey a desired order code; the designer usually needn't
consider the properties of the transistors as circuit elements.
The designer of a higher level language works with the order
code and doesn't have to know abut the logic; the user of
the higher order language needn't know the computer's order code.

	In the above cases, users of the  higher level can completely
ignore  the lower level, because the  behavior of the higher  level system is
completely determined by  the values of  the higher level  variables;
e.g.   in order to determine  the outcome of a  computer program, one
should  not have to look at the  states of flip-flops.  However, when
we ascribe mental structure to humans or goals to  society, we always
get highly  incomplete systems; the  higher level behavior  cannot be
fully predicted  from  higher  level observations  and  higher  level
"laws" even when the underlying lower level behavior is determinate.

	Besides the above philosophical reasons for ascribing mental
qualities to machines, I shall argue that in order to make machines
behave intelligently, we will have to program them to ascribe beliefs
etc. to each other and to people.

→→→→Here there will be more on machine's models of each others minds.←←←←


.SKIP TO COLUMN 1
.ONCE CENTER
TWO METHODS OF DEFINITION
.ONCE CENTER
AND THEIR APPLICATION TO MENTAL QUALITIES

	In our opinion, a major source of problems in defining mental
and other philosophical concepts is the weakness of the methods of
definition that have been %2explicitly%1 used.  We will consider two
kinds of definition: %2second order structural definition%1 and
%2definition relative to an approximate theory%1 and their application
to defining mental qualities.

.ITEM←0;

#. %3Second Order Structural Definition.%1

	Structural definitions of qualities are given in terms of the
state of the  system being described while behavioral definitions are
given in terms of actual or potential behavior$.

	~Behavioral  definitions are often  favored in  philosophy.  A
system is  defined to  have  a certain  quality if  it behaves  in  a
certain way  or  is  %2disposed%1 to behave  in a  certain way.
Their  ostensible  virtue  is  conservatism;   they  don't  postulate
internal  states that  are unobservable  to  present science  and may
remain unobservable.
However, such  definitions are awkward for  mental qualities, because,
as common sense  suggests, a  mental quality  may  not  result  in
behavior, because another mental  quality may prevent it; e.g.  I may
think you are  thick-headed, but politeness may prevent my saying so.
Particular  difficulties  can  be  overcome,  but  an  impression  of
vagueness remains.  The liking  for behavioral definitions stems from
caution, but  I would interpret scientific experience as showing that
boldness in postulating  complex structures of unobserved  entities -
provided it is  accompanied by a willingness to  take back mistakes -
is more likely to be rewarded by understanding of  and  control  over
nature   than   is   positivistic   timidity.    It  is  particularly
instructive to imagine a determined behaviorist trying to figure  out
an  electronic  computer.  Trying to define each quality behaviorally
would get him nowhere;  only  simultaneously  postulating  a  complex
structure  including  memory, arithmetic unit, control structure, and
input-output would yield predictions  that  could  be  compared  with
experiment.~

	In the case of a specific known machine, one can often give a
%2first  order  structural  definition%1.    Thus  we  might  give  a
predicate %2B(s,p)%1 such that if the  machine is in state %2s%1,  it
is said to  believe the sentence  %2p%1 provided %2B(s,p)%1  is true.
However, I don't think there is a general definition of belief having
this form that applies to all machines in all environments.

	Therefore we give a %2second order predicate%1  β%2(B,W)%1 that
tells whether we regard the  first order predicate %2B(s,p)%1 as a
"good" notion  of belief in the %2world W%1.  Such a predicate β will
be  called a  %2second  order  definition%1; it  gives  criteria  for
criticizing   an  ascription  of   a  quality   to  a  system.
Axiomatizations of  belief  give   rise  to  second   order
definitions, and we suggest that both our  common sense and scientific
usage  of not-directly-observable qualities  corresponds more closely
to second  order structural  definition than  to any kind of behavioral
definition.   It  should  be  noted that  a  second order  definition
cannot  guarantee  that  there exist  predicates  %2B%1  meeting the
criterion β or that such a %2B%1 is unique.   It can  also  turn  out
that  a quality is best defined as a member of a group of related
qualities.

	The  second   order   definition  criticizes   whole   belief
structures rather than  individual beliefs.  We can  treat individual
beliefs  by  saying  that a  system  believes  %2p%1  in state  %2s%1
provided all "reasonably good"  %2B%1's satisfy %2B(s,p)%1.  Thus  we
are distinguishing the "intersection" of the reasonably good %2B%1's.

	It  seems  to  me  that  there should  be  a  metatheorem  of
mathematical  logic asserting  that not all  second order definitions
can  be reduced  to  first  order definitions  and  further  theorems
characterizing  those  second   order  definitions  that  admit  such
reductions.  Such  technical results, if  they can be  found, may  be
helpful in  philosophy and in  the construction of  formal scientific
theories.  I would conjecture that many of the informal philosophical
arguments that certain mental  concepts cannot be reduced to  physics
will turn out to be sketches of arguments that these concepts require
second (or higher) order definitions.


	Here  is a deliberately imprecise second order definition of
belief.  For each state %2s%1 of the  machine and
each  sentence %2p%1 in  a suitable  language %2L%1,
a belief predicate %2B(s,p)%1 assigns  truth  or falsity
according to whether the machine is considered to believe %2p%1  when
it  is  in state  %2s%1.    The  language  %2L%1 is  chosen  for  our
convenience, and  there is no assumption  that the machine explicitly
represents sentences of %2L%1 in any way.  Thus we can talk about the
beliefs of Chinese, dogs, thermostats, and computer operating systems
without  assuming that they  use English or our  favorite first order
language.

	We now subject %2B(s,p)%1 to the certain criteria; i.e. β⊗(B,W) 
is considered true provided the following conditions are satisfied:

	&. The set %2Bel(s)%1 of beliefs, i.e. the set of %2p%1's for
which  %2B(s,p)%1 is  assigned true  contains  sufficiently "obvious"
consequences of some of its members.

	&.   %2Bel(s)%1 changes in a  reasonable  way when the  state
changes.    We   like  new  beliefs  to  be  logical  or  "plausible"
consequences of old ones or to come in as %2communications%1  in some
language on the input lines or  to be %2observations%1, i.e.  beliefs
about the environment the information for which comes in on the input
lines.  The set of beliefs should not change too rapidly as the state
changes.

	&.   We prefer  the set  of beliefs  to be  as consistent  as
possible.  (Admittedly, consistency  is not a quantitative concept in
mathematical logic -  a system is  either consistent or  not, but  it
would seem that we  will sometimes have to ascribe  inconsistent sets
of beliefs to machines and people.  Our intuition says that we should
be able to maintain areas of  consistency in our beliefs and that  it
may be  especially important  to avoid  inconsistencies in  the machine's
purely analytic beliefs).

	&.  Our criteria for belief systems can be strengthened if we
identify some of the  machine's beliefs as expressing goals,  i.e. if
we have beliefs  of the form "It would be good if  ...".  Then we can
ask that the machine's behavior  be somewhat %2rational%1, i.e.  %2it
does what  it  believes will  achieve its  goals%1. The  more of  its
behavior  we can  account for in  this way,  the better we  will like
%2B(s,p)%1.  We also would like to account for internal state changes
as changes in belief in so far as this is reasonable.

	&.  If the machine communicates, i.e. emits sentences in some
language  that  can  be  interpreted  as  assertions,  questions  and
commands, we will want the assertions to be among  its beliefs unless
we are  ascribing to it  a goal or  subgoal that involves  lying.  In
general, its  communications  should  be such  as  it  believes  will
achieve its goals.

	&.  Sometimes we shall want to ascribe introspective beliefs,
e.g. a belief that it does not know how to fly to Boston or even that
it doesn't know what it wants in a certain situation.

	&. Finally, we will prefer a more economical ascription %2B%1
to a less economical one.  The  fewer beliefs we ascribe and the less
they change with state consistent with accounting for he
behavior and the internal state changes, the better we will like it.

	The above  criteria have  been  formulated somewhat  vaguely.
This  would be bad if there  were widely different ascriptions of  beliefs
to a particular machine that all met our  criteria or if the criteria
allowed  ascriptions that  differed widely from  our intuitions.   My
present opinion is that more thought will make the  criteria somewhat
more precise  at no cost  in applicability, but that  they %2should%1
still remain  rather vague, i.e. we shall want to ascribe belief in a
%2family%1 of cases.
However, even at the present level of vagueness, there probably
won't be radically different equally "good" ascriptions of belief for
systems of practical interest.  If there were, we would notice the
ambiguity in ordinary language.

	Of course we will need precise axiomatizations of belief and
other mental qualities to build into particular intelligent computer programs.


#. %3Definitions relative to an approximate theory%1.


	Certain  concepts,  e.g. %2X can do Y%1, are meaningful only in
connection with a rather complex theory.   For  example,  suppose  we
denote the state of the world by %2s%1, and suppose we have functions
%2f%41%2(s)%1,...,%2f%4n%2(s)%1  that  are  directly  or   indirectly
observable.  Suppose further that %2F(s)%1 is another function of the
world-state  but  that  we   can   approximate   it by

	%2F"(s) = F'(f%41%2(s),...,f%4n%2(s))%1.

	Now  consider  the counterfactual
conditional sentence, "If %2f%42%2(s)%1 were 4, then %2F(s)%1 would be
3 - calling the present state of the world %2s%40%1." By itself, this
sentence has no meaning, because no definite state %2s%1 of the world
is  specified  by  the  condition.   However, in the framework of the
functions %2f%41%2(s),...,f%4n%2(s)%1 and the given approximation  to
%2F(s)%1,  the  assertion  can be verified by computing ⊗F' with all
arguments except the second having the  values  associated  with  the
state %2s%40%1 of the world.

	This gives rise to some remarks:

&. The most straightforward case of counterfactuals arises when
the state of a phenomenon has a distinguished Cartesian product
structure.  Then the meaning of a change of one component without
changing the others is quite clear.  Changes of more than one
component also have definite meanings.  This is a stronger
structure than the %2possible worlds%1 structure discussed in
(Lewis 1973).

&. The usual case is one in which the state %2s%1 is a substantially
unknown entity and the form of the function %2F%1 is also
unknown, but the values of %2f%41%2(s),...,f%4n%2(s)%1 and
the function %2F'%1 are much better known.
Suppose further that %2F"(s)%1 is known to be only a fair approximation
to %2F(s)%1.  We now have a situation in which the counterfactual
conditional statement is meaningful as long as it is not examined too
closely, i.e. as long as we are thinking of the world in terms of
the values of %2f%41%2,...,f%4n%1, but when we go beyond the
approximate theory, the whole meaning of the sentence seems to
disintegrate.

	Our idea is that this is a very common phenomenon. In
particular it applies to statements of the form %2"X can do Y"%1.
Such statements can be given a precise meaning in terms
of a system of interacting automata as is discussed in detail in
(McCarthy and Hayes 1970).  We say that Automaton 1 can put Automaton
3 in state 5 at time 10 by asking a question about an automaton system
in which the outputs from Automaton 1 are replaced by inputs from
outside the system.  Namely, we ask whether there is a sequence of
inputs to the new system that ⊗would put Automaton 3 in state 5 at time 10;
if yes, we say that Automaton 1 ⊗could do it in the original system
even though we may be able to show that he won't emit the necessary
outputs.  In that paper, we argue that this definition corresponds
to the intuitive notion of %2X can do Y.%1.

	What was not noted in that paper is that modelling the
situation by the particular system of interacting automata is
an approximation, and the sentences involving
⊗can derived from the approximation
cannot necessarily be translated into single assertions about the
real world. 

	I contend  that the statement,  %2"I can go  skiing tomorrow,
but I don't intend to, because I want to finish this paper"%1 has the
following properties:

	1. It has a precise meaning in a certain approximate theory of
in which I and my environment are considered as collections of interacting
automata.

	2. It cannot be directly interpreted as a statement about the world
itself, because it can't be stated in what total configurations of the
world the success of my attempt to go skiing is to be validated.

	3. The approximate theory within which the statement is meaningful
may have an objectively preferred status in that it may be the only theory
not enormously more complex that enables my actions and mental states to
be predicted.

	4. The statement may convey useful information.

Our conclusion is that the statement is %3true%1, but in a sense that
depends essentially on the approximate theory, and that this intellectual
situation is normal and should be accepted.  We further conclude that
the old-fashioned common sense analysis of a personality into %2will%1
and %2intellect%1 and other components may be valid and might be put
on a precise scientific footing using %2definitions relative to an
approximate theory%1.


.SKIP TO COLUMN 1
.ONCE CENTER
EXAMPLES OF SYSTEMS WITH MENTAL QUALITIES

	Let us  consider some  examples of machines  and programs  to
which we may ascribe belief and goal structures.

.ITEM←0;
	#.  %3Thermostats.%1   Ascribing beliefs to simple thermostats is
not really necessary, because their operation can be well  understood
without it.  However, their very simplicity  makes it clearer what is
involved  in the  ascription, and  we maintain (to  some extent  as a
provocation aimed  at  those who  regard  attribution of  beliefs  to
machines  as mere  intellectual  sloppiness) that  the ascription  is
legitimate even if unnecessary.$

~Whether a system has beliefs and other mental qualities is not
primarily a matter of complexity of the system.  Although cars are
more complex than thermostats, it is hard to ascribe beliefs or
goals to them, and the same is perhaps true of the basic hardware of a
computer, i.e. the part of the computer that executes the program
without the program itself.
~
	First  let us consider a simple  thermostat that turns off the
heat when the  temperature is a degree  above the temperature set  on
the thermostat,  turns on the heat  when the temperature  is a degree
below the desired  temperature, and leaves  the heat as  is when  the
temperature  is   in  the  two   degree  range  around   the  desired
temperature. The simplest belief predicate %2B(s,p)%1 ascribes belief
to only two sentences:  "The room is too cold"  and "The room is  too
hot", and these  beliefs are assigned to states of  the thermostat so
that  in  the  two  degree range,  neither  is  believed.    When the
thermostat believes  the  room is  too cold  or too  hot  it sends  a
message to that effect to the furnace. A slightly more complex belief
predicate could also  be used in  which the thermostat  has a  belief
about what the temperature should be and another belief about what it
is.   It is not clear  which is better, but if  we wished to consider
possible errors in the thermometer, the then we would have to be able
to ascribe beliefs  about what the temperature is. We  do not ascribe
to  it any other  beliefs; it has  no opinion even  about whether the
heat is on or off or about the weather or about who won the battle of
Waterloo.    Moreover, it  has  no introspective  beliefs,  i.e.   it
doesn't believe that it believes the room is too hot.

	The temperature control system in  my house may be  described
as  follows: Thermostats  upstairs and  downstairs  tell the  central
system  to turn  on or shut  off hot  water flow  to these areas.   A
central water-temperature thermostat tells the furnace to turn  on or
off  thus  keeping the  central  hot  water  reservoir at  the  right
temperture.  Recently it was too hot upstairs, and the question arose
as to whether the upstairs thermostat mistakenly  %2believed%1 it was
too  cold  upstairs  or  whether  the furnace  thermostat  mistakenly
%2believed %1 the  water was too  cold.  It  turned out that  neither
mistake was made; the downstairs controller %2tried%1 to turn off the
flow  of water but  %2couldn't%1, because the  valve was  stuck.  The
plumber came  once  and found  the trouble,  and  came again  when  a
replacement valve  was ordered.   Since the services of  plumbers are
increasingly  expensive, and  microcomputers are  increasingly cheap,
one is led to design a temperature control system that would %2know%1
a lot more about the thermal  state of the house and its own state of
health.

	In the first  place, while the  system %2couldn't%1 turn  off
the flow of hot water  upstairs, there is no reason  to ascribe to
it the  %2knowledge%1 that it couldn't, and  %2a fortiori%1 it had no
ability to %2communicate%1 this %2fact%1  or to take it into  account
in controlling the system.  A more advanced system would know whether
the  %2actions%1 it %2attempted%1 succeeded, and it would communicate
failures and adapt  to them.  (We  adapted to the failure  by turning
off  the whole  system  until the  whole  house cooled  off and  then
letting the two parts warm up  together.  The present system has  the
%2physical  capability%1  of  doing  this  even   if  it  hasn't  the
%2knowledge%1 or the %2will%1.


	#. %3Self-reproducing  intelligent  configurations  in  a  cellular
automaton world.%1  A ⊗cellular ⊗automaton system assigns to each vertex
in  a  certain graph a finite automaton.  The state of each automaton
at time ⊗t+1 depends on its state at time %2t%1 and the states  of
its  neighbors  at  time %2t%1. The most common graph is the array of
points ⊗(x,y) in the plane with integer co-ordinates ⊗x and ⊗y.   The
first  use of cellular automata was by von Neumann (196?) who found a
27 state automaton that could be used to  construct  self-reproducing
configuration   that  were  also  universal  computers.    The  basic
automaton in von Neumann's system had a distinguished state called  0
and  a  point in state 0 whose four neighbors were also in that state
would remain in state 0.  The initial configurations  considered  had
all  but  a  finite  number of cells in state 0, and, of course, this
property would persist although the number of  non-zero  cells  might
grow indefinitely with time.

	The  self-reproducing  system used the states of a long strip
of non-zero cells as a "tape" containing instructions to a "universal
constructor"  configuration  that  would  construct  a  copy  of  the
configuration to be reproduced but with each cell in a passive  state
that  would  persist  as  long  as its neighbors were also in passive
states.  After the construction phase, the tape would  be  copied  to
make  the  tape for the new machine, and then the new system would be
set in motion by activating one of its cells. The  new  system  would
then move away from its mother and the process would start over.  The
purpose of the design was to  demonstrate  that  arbitrarily  complex
configurations  could  be  self-reproducing  -  the  complexity being
assured by also requiring that they be universal computers.

	Since von Neumann's time, simpler basic cells admitting
self-reproducing universal computers have been discovered.  Perhaps
the simplest is the two state Life automaton of John Conway (196?).
The state of a cell at time ⊗t+1 is determined its state at time ⊗t 
and the states of its eight neighbors at time ⊗t.  Namely, if
a point whose state is 0 will change to state 1 if exactly three
of its neighbors are in state 1.  A point whose state is 1 will
remain in state 1 if two or three of its neighbors are in state 1.
In all other cases the state becomes or remains 0.

	Conway's initial intent was to model a birth and death
process whereby a cell is born (goes into state 1) if it has the
right number of living neighbors (namely three) and dies if it
is either too lonely (has none or one neighbor in state 1) or is
overcrowded (has four or more living neighbors).  He also asked
whether infinitely growing configurations were possible, and
Gosper first proved that there were.
More than that, it was shown that self-reproducing universal
computers could be built up as Life configurations.

	Let us now imagine that there are a number of such
self-reproducing universal computers operating in the Life
plane and suppose that they have been programmed to study
the properties of their world and to communicate among themselves
about it pursuing various goals co-operatively and competitively.
Let's call these configurations robots.
In some respects their intellectual and scientific problems
will be like ours, but in one major respect they will live
in a simpler world than ours has been shown to be.  Namely, 
the fundamental physics of their world is that of the life
automaton, and there is no obstacle to each robot ⊗knowing 
this physics at least in the sense of being able to simulate
the evolution of a life configuration starting in the initial
state.  Moreover, if the initial state of the robot world is finite it can
have been recorded in each robot in the beginning or else
recorded on a strip of cells that the robots can read.
(The infinite regress of having to describe the description
is avoided by a convention that the description is not
described, but can be read ⊗both as a description of the
world ⊗and as a description of the description itself.)

	These robots then know the initial state of their world
and its laws of motion.  Therefore, they can simulate as much
of their world's history as they want assuming that each of them
can grow into unoccupied space so as to have memory to store
the states of the world being simulated.  Of course, this simulation
is much slower than real time so they can never catch up with the
present let alone predict the future.  This is quite evident if we
imagine the simulation carried out in a straightforward way in which
a list of currently active cells in the simulated world is updated
according to the Life rule, but it also applies to clever mathematical
methods that might predict millions of steps ahead.  (Some Life
configurations, e.g. static ones or ones containing single ⊗gliders 
or ⊗cannon can have their distant futures predicted with little
computing.)  Namely, if there were an algorithm for such prediction,
a robot could be made that would predict its own future and then
disobey the prediction.

	Now we come to the point of this long disquisition.  Suppose
that we wish to program a robot to be successful in the Life world in
competition or co-operation with the others.  Without any idea
of how to give a mathematical proof, I will claim that our robot
will need programs that ascribe purposes and beliefs to its fellow robots
and  predict how they will react to our robot's own
actions by assuming that %2they will act in ways that they believe
will achieve their goals%1.  Our robot can acquire these mental theories
in several ways:  First, we might design in such programs and
install them in the initial configuration of the world.  Second, it might
be programmed to acquire these programs by induction from its experience
and perhaps pass them on to others through an educational system.  Third,
it might derive the psychological laws from the fundamental physics of
the world and its knowledge of the initial configuration
or it might discover how robots are built from Life cells by doing
experimental "biology".

	Knowing the Life  physics without some information  about the
initial configuration is insufficient to derive the %2psychological%1
laws, because  robots can  be constructed  in the  Life  world in  an
infinity of ways.  This follows from the "folk theorem" that the Life
automaton  is universal in the sense  that any cellular automaton can
be constructed by taking sufficiently large squares of  Life cells as
the basic cell of the other automaton.$

	Our own intellectual position is more difficult than that of
the Life robots.  We don't know the fundamental physics of our world,
and we can't even be sure that its fundamental physics is describable
in finite terms.  Even if we knew the physical laws, they seem to
preclude precise knowledge of an initial state and precise calculation
of its future both for quantum mechanical reasons and because the
continuous functions involved can't necessarily be described by
a finite amount of information.

~	Our ability to derive the laws of
higher levels of organization from knowledge of lower level laws is
also limited by universality.  Namely, while there appears to be
essentially one possible chemistry allowed by the laws of physics,
the laws of physics and chemistry allow many biologies, and, because
the neuron is a universal computing element, an arbitrary mental
structure is allowed by basic neurophysiology.  This means that in
order to determine human mental structure one must either make
psychological experiments or determine the actual anatomical structure
of the brain and the information stored in it or reason from the
fact that the brain is capable of certain problem solving performance
to the structures that must be present to provide that performance.
In this respect, our position is similar to that of the Life robot.~

	The point of the cellular automaton robot example is that much
of human mental structure is not an accident of evolution or even
of the physics of our world, but is required for successful problem
solving behavior and must be designed into or evolved by any system
that exhibits such behavior.


	#. %3Computer time-sharing systems.%1  These complicated
computer programs allocate computer time and other resources among users.
They allow each user of the computer to behave as though he had
a computer of his own, but also allow them to share files of
data and programs and to communicate with each other.
They are often used for many years with continual small changes, and
and the people making the changes and correcting errors
are often different not the original authors of the system.
A person confronted with the task of correcting a malfunction or
making a change in a time-sharing system can conveniently use a
mentalistic model of the system.

	Thus suppose a user complains that the system
will not run his program.  Perhaps the system believes that he
doesn't want to run, perhaps it persistently believes that he
has just run, perhaps it believes that his quota of computer
resources is exhausted, or perhaps it believes that his program
requires a resource that is unavailable.  Testing these hypotheses
can often be done with surprisingly little understanding of the
internal workings of the program.

→→→→→There will be more examples here of the belief of time-sharing systems.←←←

	#. %3Programs designed to reason.%1  Suppose we explicitly design a
program to represent information by sentences in a certain language
stored in the memory of the computer and decide what to do by making
inferences, and doing what it concludes will advance its goals.  Naturally,
we would hope that our previous second order definition of belief will
"approve of" a %2B(p,s)%1 that ascribed to the program believing the
sentences explicitly built in.  We would be somewhat embarassed if
someone were to show that our second order definition approved as
well or better of an entirely different set of beliefs.

	Such a program was first proposed in (McCarthy 1960), and here is how
it might work:

	Information about the world is stored in a wide variety of
data structures.  For example, a visual scene received by a TV
camera may be represented by a 512x512x3 array of numbers representing
the intensities of three colors at the points of the visual field.
At another level, the same scene may be represented by a list of regions,
and at a further level there may be a list of physical objects and their
parts together with other information about these objects obtained from
non-visual sources.  Moreover, information about how to solve various
kinds of problems may be represented by programs in some programming
language.

	However, all the above representations are subordinate to
a collection of sentences in a suitable first order language that
includes set theory.  By subordinate, we mean that there are sentences
that tell what the data structures represent and what the programs do.
New sentences can arise by a variety of processes: inference from
sentences already present, by computation from the data structures representing
observations, ...

→→→→→There will be more here about what mental qualities should be programmed.←←←

.SKIP TO COLUMN 1
.ONCE CENTER
"GLOSSARY" OF MENTAL QUALITIES

	In this section we give short "definitions" for machines of a collection
of mental qualities.  We include a number of terms which give us
difficulty with an indication of what the difficulties seem to be.

.ITEM←0;

	#. %3Actions%1.  We want to distinguish the actions of a being
from events that occur in its body and that affect the outside
world.  For example, we wish to distinguish a random twitch
from a purposeful movement.  This is not difficult %2relative
to a theory of belief that includes intentions%1.  One's purposeful
actions are those that would have been different had one's intentions
been different.  This requires that the theory of belief have sufficient
Cartesian product structure so that the counterfactual conditional
`"if its intentions had been different" is defined in the theory.
As explained in the section on definitions relative to an approximate
theory, it is not necessary that the counterfactual be given a
meaning in terms of the real world.

	#. %3Introspection and self-knowledge.%1

	We  say that  a  machine introspects  when it  comes  to have
beliefs about its own mental  state.  A simple form of  introspection
takes  place  when  a  program  determines  whether  it  has  certain
information  and if not asks for it.   Often an operating system will
compute a check  sum of itself  every few minutes  to verify that  it
hasn't been changed by a software or hardware malfunction.

	In principle,  introspection is easier  for computer programs
than for people, because the entire memory in which programs and data
are stored is available for inspection.   In fact, a computer program
can  be  made to  predict how  it  would react  to  particular inputs
provided it has enough free storage to perform the calculation.  This
situation smells of paradox, and there  is one.  Namely, if a program
could predict its own actions in less time than it takes to carry out
the action, it could refuse  to do what it has predicted  for itself.
This only  shows that self-simulation  is necessarily a  slow process,
and this is not surprising.

	However,   present    programs    do    little    interesting
introspection.   This is just  a matter of  the undeveloped  state of
artificial  intelligence; programmers  don't yet know  how to  make a
computer program look at itself in a useful way.

#.  %3Consciousness and self-consciousness%1.  In accordance with the
general  approach  of  this  paper,  a  being  is  considered
self-conscious iff it has certain beliefs about itself. However,
we must remember that beliefs are taken as sentences in our language,
and  by ascribing beliefs  we are  not asserting that  the being uses
that language directly or any other language.

	Here is  a hypothesis  arising  from artificial  intelligence
concerning the relation between language and thought.  Imagine a
person or machine  that represents  information internally in  a huge  network.
Each  node of  the  network has  references  to other  nodes  through
relations.    (If the system has  a  variable  collection of
relations, then the relations have to be represented by nodes, and
we get a symmetrical theory if we suppose that each node is connected
to a  set of pairs of other nodes).  We can imagine this structure to
have a long term part and also extremely temporary parts representing
current  %2thoughts%1.    Naturally,  each  being  has  a
its own network depending on  its own experience. A  thought is then  a
temporary  node  currently  being  referenced  by  the  mechanism  of
consciousness.  Its  meaning is determined by its references to other
nodes which  in turn  refer to  yet other  nodes.   Now consider  the
problem of communicating a thought to another being.

	Its full communication would  involve transmitting the entire
network  that  can  be  reached  from  the  given  node, and this would
ordinarily constitute the entire experience of the being.   More than
that,  it would be  necessary to  also communicate the  programs that
that take action on the basis of encountering certain nodes.  Even if
all this could be transmitted, the recipient would still have to find
equivalents  for  the  information  in  terms  of  its  own  network.
Therefore, thoughts  have to  be translated  into  a public  language
before they can be commuunicated.

	A language  is also a network  of associations
and  programs.  However, certain  of the nodes  in this network (more
accurately a %2family%1 of networks, since no two people speak precisely the same
language) are  associated with words  or set phrases.  Sometimes the
translation from thoughts to sentences is easy,
because large  parts of the  private
networks are taken from the public network, and there is an advantage
in preserving the correspondence.  However, the translation is always
approximate (in sense that still lacks a technical definition),
and some  areas of
experience  are  difficult to  translate  at  all.  Sometimes this is for
intrinsic  reasons, and  sometimes because  particular cultures don't
use language in this area.  (It is my impression that cultures differ
in the  extent to which information about  facial appearance that can
be used for recognition is  verbally transmitted).  According to  this
scheme, the "deep structure"  of a publicly expressible thought  is a
node in the public network.  It is translated into the deep structure
of a sentence as a tree  whose terminal nodes are the nodes to  which
words or set  phrases are attached.  This "deep  structure" then must
be translated into a string in a spoken or written language.

	The need to use language to express thought also applies when
we have to ascribe thoughts to other beings, since  we cannot put the
entire network into a single sentence.

→→→→→→→→→→There is more to come here about what ideas are ←←←←←←←←←←←
→→→→→→→→→→needed for self-consciousness.←←←←←←←←←←←←←←←←←←←←←←←←←←←←←

	#. %3Intentions.%1

	We may say that a machine intends to perform an action when it
believes that it will perform the action and it believes that the
action will further a goal.
However, further analysis may show that no such first order definition
in terms of belief adequately describes intentions.  In this case,
we can try a second order definition based on an axiomatization of
a predicate %2I(a,s)%1 meaning that the machine intends the action
⊗a when it is in state ⊗s.

	#. %3Free will%1

	When we program a computer  to  make
choices  intelligently  after determining its options,
examining their consequences, and deciding which
is  most  favorable or most moral or whatever, we must
program it  to  take  an  attitude  towards  its  freedom  of  choice
essentially  isomorphic  to  that  which a human must take to his own.

	We can define whether a particular action
was free or forced relative to a theory
that ascribes beliefs and within which
beings do what they believe will advance their goals.
In such a theory, action is precipitated by a belief of the form
%2I should do X now%1.  We will say that the action was free if
changing the belief to %2I shouldn't do X now%1 would have resulted
in the action not being performed.
This requires that the theory of belief have sufficient Cartesian
product structure so that changing a single belief is defined, but it
doesn't require defining what the state of the world would be if
a single belief were different.

	This isn't the  whole free will story, because  moralists are
also  concerned with whether praise  or blame may be  attributed to a
choice.   The following  considerations would  seem to  apply to  any
attempt to define  the morality of actions in a  way that would apply
to machines:

&. There is unlikely  to be a simple behavioral  definition.  Instead
there would be  a second order definition criticizing predicates that
ascribe morality to actions.


&. The theory must contain at least one axiom of morality that is not
just a statement of physical fact.  Relative to this axiom, judgments
of actions can be factual.

&. The theory of morality will presuppose a theory of belief in which
statements of the form %2"It believed the action would harm someone"%1
are defined.  The theory must ascribe beliefs about others' welfare and
perhaps about the being's own welfare.

&. It might be necessary to consider the  machine as imbedded in some
kind of society in order to ascribe morality to its actions.

&. No  present machines admit  such a belief  structure, and  no such
structure  may be required  to make  a machine with  arbitrarily high
intelligence in the sense of problem-solving ability.

&. It seems  doubtful that morally  judgable machines or  machines to
which rights might legitimately be ascribed are desirable if and when
it becomes possible to make them.

→→→→→→More mental qualities will be discussed.←←←←←←←←←

.SKIP TO COLUMN 1
.ONCE CENTER
OTHER VIEWS ABOUT MIND

→→→→→This section will be written←←←←←
.SKIP TO COLUMN 1
.PORTION NOTES
.ONCE CENTER
NOTES

.RECEIVE;

.ITEM←NOTE;
	#. Philosophy and artificial intelligence.  These fields overlap
in the following way:  In order to make a computer program behave
intelligently, its designer must build into it a view of the world
in general, apart from what they include about
particular sciences.  (The skeptic who doubts
whether there is anything to say about the world apart from the
particular sciences should try to write a computer program that can
figure out how to get to Timbuktoo, taking into account not only
the facts about travel in general but also facts about what people
and documents  have what information, and what information will
be required at different stages of the trip and when and how it is
to be obtained.  He will rapidly discover that he is lacking a %2science
of common sense%1, i.e. he will be unable to formally express and
build into his program "what everybody knows".  Maybe philosophy
could be defined as an attempted %2science of common sense%1,
or else the %2science of common sense%1 should be a definite part
of philosophy.)

	Artificial intelligence has a another component in which
philosophers have not studied, namely %2heuristics%1.  Heuristics
is  concerned  with:  given  the  facts  and  a  goal,  how should it
investigate the possibilities and decide what to do.
On the other hand, artificial intelligence is not much concerned
with aesthetics and ethics.

	Not all approaches to philosophy lead to results relevant to
the artificial intelligence problem.  On the face of it, a philosophy
that entailed the view that artificial intelligence was impossible
would be unhelpful, but besides that, taking artificial intelligence
seriously suggests some philosophical points of view.  I am not sure
that all I shall list are required for pursuing the AI goal -
some of them may be just my prejudices - but here they are:

		&. The relation between a world view and the world
should be studied by methods akin to metamathematics in which
systems are studied from the outside.  In metamathematics we study
the relation between a mathematical system and its models.  Philosophy
(or perhaps %2metaphilosophy%1) should study the relation between
world structures and systems within them that seek knowledge.
Just as the metamathematician can use any mathematical methods
in this study and distinguishes the methods he uses form those
being studied, so the philosopher should use all his scientific
knowledge in studying philosphical systems from the outside.

	Thus the question %2"How do I know?"%1 is best answered by studying
%2"How does it know"%1, getting the best answer that the current state
of science and philosophy permits, and then seeing how this answer stands
up to doubts about one's own sources of knowledge.

	&. We  regard  %2metaphysics%1 as  the study  of the  general
structure  of   the  world  and  %2epistemology%1  as  studying  what
knowledge of  the world  can be  had by  an  intelligence with  given
opportunities  to observe  and experiment.   We  need to  distinguish
between  what can  be determined  about the  structure of  humans and
machines  by  scientific   research  over  a   period  of  time   and
experimenting with many individuals, and  what can be learned by in a
particular situation with particular opportunities to observe.   From
the AI point of view, the latter is as important
as the  former, and we suppose that  philosophers would also consider
it part of epistemology.  The possibilities of reductionism  are also
different  for  theoretical  and  everyday epistemology.    We  could
imagine that the rules of everyday epistemology could be deduced from
a knowledge of physics and the structure of the  being and the world,
but  we  can't see  how  one could  avoid  using  mental concepts  in
expressing knowledge actually obtained by the senses.

	&.  It  is  now  accepted that the basic concepts of physical
theories are far removed from observation.  The  human  sense  organs
are  many  levels  of  organization  removed  from quantum mechanical
states, and we have learned to accept the complication this causes in
verifying physical theories. Experience in trying to make intelligent
computer programs suggests that the  basic  concepts  of  the  common
sense  world  are  also complex and not always directly accessible to
observation.   In  particular,  the  common  sense  world  is  not  a
construct  from  sense  data,  but sense data play an important role.
When a man or a computer program sees a dog, we will  need  both  the
relation  between  the  observer and the dog and the relation between
the observer and the brown patch in order to construct a good  theory
of the event.

	&.  In spirit this paper is  materialist, but it is logically
compatible  with some  other philosophies.   Thus  cellular automaton
models of the  physical world may be  supplemented by supposing  that
certain  complex  configurations  interact with  additional  automata
called   souls   that  also   interact   with  each   other.     Such
%2interactionist  dualism%1   won't  meet   emotional  or   spiritual
objections to  materialism, but it does provide a logical  niche for any
empirically  argued  belief in  telepathy, communication with the dead
and  other   psychic phenomena.
A person who believed the alleged evidence for such phenomena and
still wanted a scientific explanation could model his beliefs
with auxiliary automata.

.SKIP TO COLUMN 1
.ONCE CENTER
REFERENCES

→→→→→→→→References will be supplied←←←←←←←←
.BEGIN VERBATIM

John McCarthy
Artificial Intelligence Laboratory
Stanford University
Stanford, California 94305
.END